Robust BCa-JaB Method as a Diagnostic Tool for Linear Regression Models
نویسنده
چکیده
The detection and evaluation of influential observations are critical aspects of data analysis in the context of linear regression models. The assessment of whether the change in the model resulting from the inclusion/exclusion of each respective data point is major is usually based on cut-off values for the influence measure obtained from asymptotic approximations. However, the asymptotic approximations used for these diagnostic measures suffer both because the null distributions of these quantities are very complex and because the approximations tend to be poor when sample sizes are small. As computing power has exploded over the intervening decades, computerintensive methods such as Jackknife-after-Bootstrap (JaB) technique has been proposed by various authors as a method for detecting influence in regression models through refining the cut-off values for the common influence diagnostics used in linear regression models. The rationale behind the use of the JaB method in this context is its construction based on deleting observations one at a time, with cut-offs based on the bootstrap distribution of the influence measure calculated using the JaB approach which allows the consideration of bootstrap samples that are free of the influence of a particular data point. The method is based on the use of percentile-method bootstrap confidence intervals which have well-known deficiencies under skewed distributions. In the present context, using this method will manifest as a tendency to fail to flag points that may, in fact, be unusual. In order to improve JaB, we propose using robust versions of bias-corrected and accelerated (BCa) bootstrap confidence intervals which tolerate skewed distributions. We show that, asymptotically, proposed equal-tailed two-sided interval is second-order accurate. The performances of the proposed robust BCa-JaB and conventional JaB methods are compared in the cases of DFFITS, Welsch’s distance and modified Cook’s distance influence diagnostics. In this talk, I will present the results for both real data examples and simulation study. To request an interpreter or other accommodations for people with disabilities, please call the Department of Statistics and Probability at 517-355-9589.
منابع مشابه
Jackknife-After-Bootstrap as Logistic Regression Diagnostic Tool
Jackknife-after-Bootstrap (JaB) has first been proposed by [1] then used by [2] and [3] to detect influential observations in linear regression models. In this study, we propose using JaB to detect influential observations in logistic regression model. Performance of the proposed method will be compared with the traditional method for standardized Pearson residuals, Cook’s distance, change in t...
متن کاملRobust Estimation in Linear Regression with Molticollinearity and Sparse Models
One of the factors affecting the statistical analysis of the data is the presence of outliers. The methods which are not affected by the outliers are called robust methods. Robust regression methods are robust estimation methods of regression model parameters in the presence of outliers. Besides outliers, the linear dependency of regressor variables, which is called multicollinearity...
متن کاملRobust high-dimensional semiparametric regression using optimized differencing method applied to the vitamin B2 production data
Background and purpose: By evolving science, knowledge, and technology, we deal with high-dimensional data in which the number of predictors may considerably exceed the sample size. The main problems with high-dimensional data are the estimation of the coefficients and interpretation. For high-dimension problems, classical methods are not reliable because of a large number of predictor variable...
متن کاملAnalysis of a Problem Using Various Visions
In this paper an applied problem, where the response of interest is the number of success in a specific experiment, is considered and by various visions is studied. The effects of outlier values of response on results of a regression analysis are so important to be studied. For this reason, using diagnostic methods, outlier response values are recognized. It is shown that use of arc-sine ...
متن کاملRobust Estimation in Linear Regression Model: the Density Power Divergence Approach
The minimum density power divergence method provides a robust estimate in the face of a situation where the dataset includes a number of outlier data. In this study, we introduce and use a robust minimum density power divergence estimator to estimate the parameters of the linear regression model and then with some numerical examples of linear regression model, we show the robustness of this est...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2014